منابع مشابه
Visualized voices: a case study of audio-visual synesthesia.
We report a single case study of a synesthete (PS) who has complex visual experiences from sounds, including human voices. Different vowel sounds from different speakers and modified to be of different pitch (f0) were presented to PS and controls who were asked to draw an (abstract) visual image of the sound noting colors, sizes, and locations. PS tended to be more consistent over time than con...
متن کاملAn Audio-Haptic Aesthetic Framework Influenced by Visual Theory
Sound is touch at a distance. The vibration of pressure waves in the air creates sounds that our ears hear, at close range, these pressure waves may also be felt as vibration. This audio-haptic relationship has potential for enriching interaction in human-computer interfaces. How can interface designers manipulate attention using audio-haptic media? We propose a theoretical perceptual framework...
متن کاملHandling large audio files in audio books for building synthetic voices
One of the issues in using audio books for building a synthetic voice is the segmentation of large audio files. The use of standard forced-alignment to obtain phone boundaries on large audio files fails primarily because of huge memory requirements. Earlier works have attempted to resolve this problem by using large vocabulary speech recognition system employing restricted dictionary and langua...
متن کاملAudio-visual Person Veriication Audio-visual Person Veriication
In this paper we investigate bene ts of classi er combination fusion for a multimodal system for personal identity veri cation The system uses frontal face images and speech We show that a sophisticated fusion strategy enables the system to outperform its facial and vocal modules when taken seperately We show that both trained linear weighted schemes and fusion by Support Vector Machine classi ...
متن کاملContinuous Audio-visual Speech Recognition Continuous Audio-visual Speech Recognition
We address the problem of robust lip tracking, visual speech feature extraction, and sensor integration for audiovisual speech recognition applications. An appearance based model of the articulators, which represents linguistically important features, is learned from example images and is used to locate, track, and recover visual speech information. We tackle the problem of joint temporal model...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: InMedia
سال: 2013
ISSN: 2259-4728
DOI: 10.4000/inmedia.697